Goto

Collaborating Authors

 probabilistic perspective


Concept-based Adversarial Attack: a Probabilistic Perspective

Zhang, Andi, Ding, Xuan, McDonagh, Steven, Kaski, Samuel

arXiv.org Artificial Intelligence

We propose a concept-based adversarial attack framework that extends beyond single-image perturbations by adopting a probabilistic perspective. Rather than modifying a single image, our method operates on an entire concept -- represented by a probabilistic generative model or a set of images -- to generate diverse adversarial examples. Preserving the concept is essential, as it ensures that the resulting adversarial images remain identifiable as instances of the original underlying category or identity. By sampling from this concept-based adversarial distribution, we generate images that maintain the original concept but vary in pose, viewpoint, or background, thereby misleading the classifier. Mathematically, this framework remains consistent with traditional adversarial attacks in a principled manner. Our theoretical and empirical results demonstrate that concept-based adversarial attacks yield more diverse adversarial examples and effectively preserve the underlying concept, while achieving higher attack efficiency.


Bayesian Deep Learning and a Probabilistic Perspective of Generalization

Neural Information Processing Systems

The key distinguishing property of a Bayesian approach is marginalization, rather than using a single setting of weights. Bayesian marginalization can particularly improve the accuracy and calibration of modern deep neural networks, which are typically underspecified by the data, and can represent many compelling but different solutions. We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization, and propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction, without significant overhead. We also investigate the prior over functions implied by a vague distribution over neural network weights, explaining the generalization properties of such models from a probabilistic perspective. From this perspective, we explain results that have been presented as mysterious and distinct to neural network generalization, such as the ability to fit images with random labels, and show that these results can be reproduced with Gaussian processes.


Review for NeurIPS paper: Bayesian Deep Learning and a Probabilistic Perspective of Generalization

Neural Information Processing Systems

Summary and Contributions: This paper provides a mix between discussing high-level conceptual ideas and perspectives and presenting a variety of experimental results, all under the umbrella of generalization in (Bayesian) deep learning. More concretely, the central argument of the paper is that Bayesian learning should be primarily viewed as aiming to marginalize over different plausible hypotheses of the data, intead of relying on a single hypothesis (which is what ordinary deep learning is doing). The ultimate goal is thus to accurately estimate the posterior _predictive_ distribution (over outputs), rather than to accurately approximate the posterior distribution (over weights). They thus recommend that Bayesian methods should ideally focus their efforts on carefully representing the posterior distribution in regions that contribute most to the predictive distribution. In this line of thought, they further argue that deep ensembles, one of the state-of-the-art approaches for obtaining well-calibrated predictive distributions, do effectively approximate the Bayesian model average (even if the individual ensemble members are not actually samples from the posterior), and thus should not be considered in competition to Bayesian methods.


Review for NeurIPS paper: Bayesian Deep Learning and a Probabilistic Perspective of Generalization

Neural Information Processing Systems

After much discussion, the reviewers largely converged towards recommending to accept this submission. The reviewers appreciate the merits of the paper, believe it investigates important open questions, and will thus be a significant contribution to our understanding of BNNs, but only when the experimental issues mentioned in the reviews are resolved. I would draw the author's attention to the fact that the reviewers raised concerns about the supplementary material containing a number of sections which are not connected to results in the main paper (on tempered posteriors, sampling from the prior, discussions of what's Bayesian, PAC Bayes etc.). Per reviewing guidelines, since these sections were not relevant for understanding the main paper, these were not reviewed with scrutiny. However, the reviewers found strong statements in the unreviewed supplementary material involving other recent work which they believe deserve close scrutiny if they are to be published.


Reviews: High resolution neural connectivity from incomplete tracing data using nonnegative spline regression

Neural Information Processing Systems

The underlying model is a non-negative linear regression, y Wx \eta, where \eta is drawn from a spherical Gaussian model. The weight matrix, W, is assumed to be nonnegative and, in probabilistic terms, drawn from a spatially smooth prior. Optionally, a low-rank assumption may be incorporated into the weight model, which can dramatically improve memory efficiency for large-scale problems. While the individual components of this model (nonnegative regression, Laplacian regularized least squares, low-rank constraints) are well-studied, I think this is a nice combination and application of these techniques to a real-world, scientific problem. The presentation of the model, the synthetic examples, and the real world applications (and supplementary movies) are particularly clear. While it is certainly valid to directly construct a objective function that captures both the reconstruction error and the domain-specific constraints and inductive biases, I think a probabilistic perspective could elucidate a number of potential extensions and connections to existing work.


Bayesian Deep Learning and a Probabilistic Perspective of Generalization

Neural Information Processing Systems

The key distinguishing property of a Bayesian approach is marginalization, rather than using a single setting of weights. Bayesian marginalization can particularly improve the accuracy and calibration of modern deep neural networks, which are typically underspecified by the data, and can represent many compelling but different solutions. We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization, and propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction, without significant overhead. We also investigate the prior over functions implied by a vague distribution over neural network weights, explaining the generalization properties of such models from a probabilistic perspective. From this perspective, we explain results that have been presented as mysterious and distinct to neural network generalization, such as the ability to fit images with random labels, and show that these results can be reproduced with Gaussian processes.


Fairness with Overlapping Groups; a Probabilistic Perspective

Neural Information Processing Systems

In algorithmically fair prediction problems, a standard goal is to ensure the equality of fairness metrics across multiple overlapping groups simultaneously. We reconsider this standard fair classification problem using a probabilistic population analysis, which, in turn, reveals the Bayes-optimal classifier. Our approach unifies a variety of existing group-fair classification methods and enables extensions to a wide range of non-decomposable multiclass performance metrics and fairness measures. On a variety of real datasets, the proposed approach outperforms baselines in terms of its fairness-performance tradeoff.


Towards One Model for Classical Dimensionality Reduction: A Probabilistic Perspective on UMAP and t-SNE

Ravuri, Aditya, Lawrence, Neil D.

arXiv.org Machine Learning

This paper shows that the dimensionality reduction methods, UMAP and t-SNE, can be approximately recast as MAP inference methods corresponding to a generalized Wishart-based model introduced in ProbDR. This interpretation offers deeper theoretical insights into these algorithms, while introducing tools with which similar dimensionality reduction methods can be studied.


The Elements of Differentiable Programming

Blondel, Mathieu, Roulet, Vincent

arXiv.org Artificial Intelligence

Artificial intelligence has recently experienced remarkable advances, fueled by large models, vast datasets, accelerated hardware, and, last but not least, the transformative power of differentiable programming. This new programming paradigm enables end-to-end differentiation of complex computer programs (including those with control flows and data structures), making gradient-based optimization of program parameters possible. As an emerging paradigm, differentiable programming builds upon several areas of computer science and applied mathematics, including automatic differentiation, graphical models, optimization and statistics. This book presents a comprehensive review of the fundamental concepts useful for differentiable programming. We adopt two main perspectives, that of optimization and that of probability, with clear analogies between the two. Differentiable programming is not merely the differentiation of programs, but also the thoughtful design of programs intended for differentiation. By making programs differentiable, we inherently introduce probability distributions over their execution, providing a means to quantify the uncertainty associated with program outputs.


Constructing Semantics-Aware Adversarial Examples with Probabilistic Perspective

Zhang, Andi, Wischik, Damon

arXiv.org Artificial Intelligence

In this study, we introduce a novel, probabilistic viewpoint on adversarial examples, achieved through box-constrained Langevin Monte Carlo (LMC). Proceeding from this perspective, we develop an innovative approach for generating semantics-aware adversarial examples in a principled manner. This methodology transcends the restriction imposed by geometric distance, instead opting for semantic constraints. Our approach empowers individuals to incorporate their personal comprehension of semantics into the model. Through human evaluation, we validate that our semantics-aware adversarial examples maintain their inherent meaning. Experimental findings on the MNIST and SVHN datasets demonstrate that our semantics-aware adversarial examples can effectively circumvent robust adversarial training methods tailored for traditional adversarial attacks.